3 research outputs found

    Database query optimisation based on measures of regret

    Get PDF
    The query optimiser in a database management system (DBMS) is responsible for �nding a good order in which to execute the operators in a given query. However, in practice the query optimiser does not usually guarantee to �nd the best plan. This is often due to the non-availability of precise statistical data or inaccurate assumptions made by the optimiser. In this thesis we propose a robust approach to logical query optimisation that takes into account the unreliability in database statistics during the optimisation process. In particular, we study the ordering problem for selection operators and for join operators, where selectivities are modelled as intervals rather than exact values. As a measure of optimality, we use a concept from decision theory called minmax regret optimisation (MRO). When using interval selectivities, the decision problem for selection operator ordering turns out to be NP-hard. After investigating properties of the problem and identifying special cases which can be solved in polynomial time, we develop a novel heuristic for solving the general selection ordering problem in polynomial time. Experimental evaluation of the heuristic using synthetic data, the Star Schema Benchmark and real-world data sets shows that it outperforms other heuristics (which take an optimistic, pessimistic or midpoint approach) and also produces plans whose regret is on average very close to optimal. The general join ordering problem is known to be NP-hard, even for exact selectivities. So, for interval selectivities, we restrict our investigation to sets of join operators which form a chain and to plans that correspond to left-deep join trees. We investigate properties of the problem and use these, along with ideas from the selection ordering heuristic and other algorithms in the literature, to develop a polynomial-time heuristic tailored for the join ordering problem. Experimental evaluation of the heuristic shows that, once again, it performs better than the optimistic, pessimistic and midpoint heuristics. In addition, the results show that the heuristic produces plans whose regret is on average even closer to the optimal than for selection ordering

    Database query optimisation based on measures of regret

    Get PDF
    The query optimiser in a database management system (DBMS) is responsible for �nding a good order in which to execute the operators in a given query. However, in practice the query optimiser does not usually guarantee to �nd the best plan. This is often due to the non-availability of precise statistical data or inaccurate assumptions made by the optimiser. In this thesis we propose a robust approach to logical query optimisation that takes into account the unreliability in database statistics during the optimisation process. In particular, we study the ordering problem for selection operators and for join operators, where selectivities are modelled as intervals rather than exact values. As a measure of optimality, we use a concept from decision theory called minmax regret optimisation (MRO). When using interval selectivities, the decision problem for selection operator ordering turns out to be NP-hard. After investigating properties of the problem and identifying special cases which can be solved in polynomial time, we develop a novel heuristic for solving the general selection ordering problem in polynomial time. Experimental evaluation of the heuristic using synthetic data, the Star Schema Benchmark and real-world data sets shows that it outperforms other heuristics (which take an optimistic, pessimistic or midpoint approach) and also produces plans whose regret is on average very close to optimal. The general join ordering problem is known to be NP-hard, even for exact selectivities. So, for interval selectivities, we restrict our investigation to sets of join operators which form a chain and to plans that correspond to left-deep join trees. We investigate properties of the problem and use these, along with ideas from the selection ordering heuristic and other algorithms in the literature, to develop a polynomial-time heuristic tailored for the join ordering problem. Experimental evaluation of the heuristic shows that, once again, it performs better than the optimistic, pessimistic and midpoint heuristics. In addition, the results show that the heuristic produces plans whose regret is on average even closer to the optimal than for selection ordering

    Capuchin Search Algorithm With Deep Learning-Based Data Edge Verification for Blockchain-Assisted IoT Environment

    No full text
    Internet of Things (IoT) devices generate enormous quantities of data, and ensuring the authenticity and integrity of this data is essential. Blockchain (BC) serves as a transparent and secure ledger for recording each data transaction, which prevents unauthorized modification and provides a trust layer for the IoT ecosystem. Data edge verification for BC-enabled IoT platforms is a fundamental aspect of ensuring data trustworthiness, integrity, and security in the system. In such an environment, IoT devices generate vast amounts of data, and BC technology can be used to create a decentralized and immutable ledger that records all the transactions and data exchanges Machine learning (ML) in the context of BC and the IoT offers a powerful toolkit for optimizing and securing these technologies. It enables the analysis of vast and complex data generated by IoT devices, allowing for predictive maintenance, anomaly detection, and pattern recognition. ML also enhances security by developing intrusion detection systems and supporting smart contracts in BC networks. This article introduces a new Capuchin Search Algorithm with a Deep Learning based Data Edge Verification model (CSADL-DEVM) for the Blockchain assisted IoT platform. The purpose of the CSADL-DEVM technique is to integrate the BC with DL and hyperparameter tuning concepts for data edge verification in the IoT environment. In addition, IoT devices comprise a considerable level of decentralized decision-making ability, which accomplishes a consensus on the performance of intra-block transactions. In addition, the CSADL-DEVM technique applies the Elman recurrent neural network (ERNN) model for the identification and classification of faults. Moreover, the hyperparameters related to the ERNN model can be adjusted by the use of CSA which in turn better the detection results. A comprehensive set of simulations has been conducted to exhibit the enhanced results of the CSADL-DEVM method. The obtained outcomes exhibit the superior outcome of the CSADL-DEVM algorithm
    corecore